Coordinated Plan Management Using Multiagent MDPs
نویسندگان
چکیده
For the past several years, we have been developing multiagent technology to help humans coordinate their activities in complex, dynamic environments. In recent work on the DARPA COORDINATORs program, we have developed multi-agent Markov-decision process (MDP) techniques for distributed plan management. The COORDINATORs problems arrive in distributed form, with different agents getting local views of their portion of the problem and its relationship to others. Even so, the individual agents’ MDPs that capture their local planning and scheduling problem can be too large to enumerate and solve. Furthermore, the COORDINATOR agents must build and execute their plans in realtime, interacting with a world simulation that makes their actions have uncertain outcome. Accordingly, we have developed an embedded agent system that negotiates to try to find approximately-optimal distributed policies within tight time constraints. Our work draws together and extends ideas in multi-agent Markov decision processes, real-time computing, negotiation, metalevel control, and distributed constraint optimization. Contributions of our work include “unrolling” techniques for translating local hierarchical task networks to MDPs, “informed” heuristic search control of the unrolling process, and negotiation methods for allocating responsibilities across cooperating agents and using those allocations to influence local policy construction. In the rest of this paper, we describe our approach in more detail. We begin by summarizing the challenges in distributed plan management embodied in the COORDINATORs problem, and the TÆMS representation used to model the actions and interactions requiring coordination. We then de-
منابع مشابه
A POMDP-based Model for Optimizing Communication in Multiagent Systems
In this paper we address the problem of planning in multiagent systems in which the interaction between the different agents is sparse and mediated by communication. We include the process of communication explicitly as part of the decision process and illustrate how this single-agent model can be used to plan for communication. We also use the single-agent model to plan in the multiagent scena...
متن کاملSafe Stochastic Planning: Planning to Avoid Fatal States
Markov decision processes (MDPs) are applied as a standard model in Artificial Intelligence planning. MDPs are used to construct optimal or near optimal policies or plans. One area that is often missing from discussions of planning under stochastic environment is how MDPs handle safety constraints expressed as probability of reaching threat states. We introduce a method for finding a value opti...
متن کاملHeuristic Planning for Decentralized MDPs with Sparse Interactions
In this work, we explore how local interactions can simplify the process of decision-making in multiagent systems, particularly in multirobot problems. We review a recent decision-theoretic model for multiagent systems, the decentralized sparse-interaction Markov decision process (Dec-SIMDP), that explicitly distinguishes the situations in which the agents in the team must coordinate from those...
متن کاملA Scheme for Agent Collaboration in Open Multiagent Environments
In multiagent planning, an agent sometimes needs to collaborate with others to construct complex plans, or to accomplish large organizational tasks which it cannot do alone. Since each agent in a group may have incorrect beliefs about the world and incomplete knowledge, and because agent's abilities differ, constructing a coordinated collaborative plan among agents is a difficult proposition. I...
متن کاملDecentralized coordination via task decomposition and reward shaping
In this work, we introduce a method for decentralized coordination in cooperative multiagent multi-task problems where the subtasks and agents are homogeneous. Using the method proposed, the agents cooperate at the high level task selection using the knowledge they gather by learning subtasks. We introduce a subtask selection method for single agent multi-task MDPs and we extend the work to mul...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2006